1. Гуржов Н. А., Переверзев А. Л., Силантьев А. М., Примаков Е. В., Солодовников А. П. Разработка реконфигурируемого систолического массива для ускорения вычисления слоев нейронной сети. Тенденции развития науки и образования. 2024;(116-19):75–80. https://doi.org/10.18411/trnio-12-2024-868. EDN: KSKOWJ.
Gurzhov N. A., Pereverzev A. L., Silant’yev A. M., Primakov E. V., Solodovnikov A. P. Development of reconfigurable systolic array for speeding up calculation of neural network layers. Tendentsii razvitiya nauki i obrazovaniya. 2024;(116-19):75–80. (In Russ.). https://doi.org/10.18411/trnio-12-2024-868
2. Kung H. T. Why systolic architectures? Computer. 1982;15(1):37–46. https://doi.org/10.1109/MC.1982.1653825
3. TPU architecture. Google Cloud TPU. 2025. Available at: https://cloud.google.com/tpu/docs/system-architecture-tpu-vm (accessed: 18.11.2025).
4. Zehendner E. 16.1. Basic concepts of systolic systems. In: Algorithms of Informatics. Vol. 2: Applications. Ed. A. Iványi. Budapest: mondAt Kiadó; 2007, pp. 752–760.
5. Understanding matrix multiplication on a weight-stationary systolic architecture. Telesens. 30.07.2018. Available at: https://telesens.co/2018/07/30/systolic-architectures/ (accessed: 18.11.2025).
6. Benoit J., Kligys S., Chen B., Zhu M., Tang M., Howard A. et al. Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE; 2018, pp. 2704–2713. https://doi.org/10.1109/CVPR.2018.00286
7. Nagel M., Fournarakis M., Amjad R. A., Bondarenko Ye., Baalen M. van, Blankevoort T. A white paper on neural network quantization. arXiv.org. 15.06.2021. Available at: https://arxiv.org/abs/2106.08295 (accessed: 18.03.2025). https://doi.org/10.48550/arxiv.2106.08295
8. Sun W., Liu D., Zou Zh., Sun Wen., Kang Y., Chen S. Sense: Model hardware co-design for accelerating sparse CNN on systolic array. arXiv.org. 01.02.2022. Available at: https://arxiv.org/abs/2202.00389 (accessed: 18.03.2025). https://doi.org/10.48550/arxiv.2202.00389
9. Marino K., Zhang P., Prasanna V. ME-ViT: A single-load memory-efficient FPGA accelerator for vision transformers. arXiv.org. 15.02.2024. Available at: https://arxiv.org/abs/2402.09709 (accessed: 18.03.2025). https://doi.org/10.48550/arxiv.2402.09709
10. Selvam S., Ganesan V., Kumar P. FuSeConv: Fully separable convolutions for fast inference on systolic arrays. arXiv.org. 27.05.2021. Available at: https://arxiv.org/abs/2105.13434 (accessed: 18.03.2025). https://doi.org/10.48550/arxiv.2105.13434
11. Initializing weights for the convolutional and fully connected layers. Telesens. 09.04.2018. Available at: https://www.telesens.co/2018/04/09/initializing-weights-for-the-convolutional-and-fully-connected-la... (accessed: 18.11.2025).